Molecular contrastive learning of representations via graph neural networks
نویسندگان
چکیده
Molecular Machine Learning (ML) bears promise for efficient molecule property prediction and drug discovery. However, labeled data can be expensive time-consuming to acquire. Due the limited data, it is a great challenge supervised-learning ML models generalize giant chemical space. In this work, we present MolCLR: Contrastive of Representations via Graph Neural Networks (GNNs), self-supervised learning framework that leverages large unlabeled (~10M unique molecules). MolCLR pre-training, build graphs develop GNN encoders learn differentiable representations. Three graph augmentations are proposed: atom masking, bond deletion, subgraph removal. A contrastive estimator maximizes agreement from same while minimizing different molecules. Experiments show our significantly improves performance GNNs on various molecular benchmarks including both classification regression tasks. Benefiting pre-training database, even achieves state-of-the-art several challenging after fine-tuning. Additionally, further investigations demonstrate learns embed molecules into representations distinguish chemically reasonable similarities.
منابع مشابه
Deep Neural Networks for Learning Graph Representations
In this paper, we propose a novel model for learning graph representations, which generates a low-dimensional vector representation for each vertex by capturing the graph structural information. Different from other previous research efforts, we adopt a random surfing model to capture graph structural information directly, instead of using the samplingbased method for generating linear sequence...
متن کاملContrastive Learning and Neural Oscillations
The concept of Contrastive Learning (CL) is developed as a family of possible learning algorithms for neural networks. CL is an extension of Deterministic Boltzmann Machines to more general dynamical systems. During learning, the network oscillates between two phases. One phase has a teacher signal and one phase has no teacher signal. The weights are updated using a learning rule that correspon...
متن کاملLearning Anonymized Representations with Adversarial Neural Networks
Statistical methods protecting sensitive information or the identity of the data owner have become critical to ensure privacy of individuals as well as of organizations. This paper investigates anonymization methods based on representation learning and deep neural networks, and motivated by novel informationtheoretical bounds. We introduce a novel training objective for simultaneously training ...
متن کاملGraph Convolutional Neural Networks via Scattering
We generalize the scattering transform to graphs and consequently construct a convolutional neural network on graphs. We show that under certain conditions, any feature generated by such a network is approximately invariant to permutations and stable to graph manipulations. Numerical results demonstrate competitive performance on relevant datasets.
متن کاملFew-Shot Learning with Graph Neural Networks
We propose to study the problem of few-shot learning with the prism of inference on a partially observed graphical model, constructed from a collection of input images whose label can be either observed or not. By assimilating generic message-passing inference algorithms with their neural-network counterparts, we define a graph neural network architecture that generalizes several of the recentl...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Nature Machine Intelligence
سال: 2022
ISSN: ['2522-5839']
DOI: https://doi.org/10.1038/s42256-022-00447-x